On stochastic gradient Langevin dynamics with dependent data streams in the logconcave case
نویسندگان
چکیده
We study the problem of sampling from a probability distribution $\pi $ on $\mathbb{R}^{d}$ which has density w.r.t. Lebesgue measure known up to normalization factor $x\mapsto \mathrm{e}^{-U(x)}/\int _{\mathbb{R}^{d}}\mathrm{e}^{-U(y)}\,\mathrm{d}y$. analyze method based Euler discretization Langevin stochastic differential equations under assumptions that potential $U$ is continuously differentiable, $\nabla U$ Lipschitz, and strongly concave. focus case where gradient log-density cannot be directly computed but unbiased estimates possibly dependent observations are available. This setting can seen as combination approximation (here gradient) type algorithms with discretized dynamics. obtain an upper bound Wasserstein-2 distance between law iterates this algorithm target constants depending explicitly Lipschitz strong convexity dimension space. Finally, weaker its in presence independent observations, we analogous results distance.
منابع مشابه
Consistency and fluctuations for stochastic gradient Langevin dynamics Consistency and fluctuations for stochastic gradient Langevin dynamics
Applying standard Markov chain Monte Carlo (MCMC) algorithms to large data sets is computationally expensive. Both the calculation of the acceptance probability and the creation of informed proposals usually require an iteration through the whole data set. The recently proposed stochastic gradient Langevin dynamics (SGLD) method circumvents this problem by generating proposals which are only ba...
متن کاملVariance Reduction in Stochastic Gradient Langevin Dynamics
Stochastic gradient-based Monte Carlo methods such as stochastic gradient Langevin dynamics are useful tools for posterior inference on large scale datasets in many machine learning applications. These methods scale to large datasets by using noisy gradients calculated using a mini-batch or subset of the dataset. However, the high variance inherent in these noisy gradients degrades performance ...
متن کاملStochastic Gradient Riemannian Langevin Dynamics on the Probability Simplex
In this paper we investigate the use of Langevin Monte Carlo methods on the probability simplex and propose a new method, Stochastic gradient Riemannian Langevin dynamics, which is simple to implement and can be applied to large scale data. We apply this method to latent Dirichlet allocation in an online minibatch setting, and demonstrate that it achieves substantial performance improvements ov...
متن کاملBayesian Learning via Stochastic Gradient Langevin Dynamics
In this paper we propose a new framework for learning from large scale datasets based on iterative learning from small mini-batches. By adding the right amount of noise to a standard stochastic gradient optimization algorithm we show that the iterates will converge to samples from the true posterior distribution as we anneal the stepsize. This seamless transition between optimization and Bayesi...
متن کاملGradient Boosting on Stochastic Data Streams
Boosting is a popular ensemble algorithm that generates more powerful learners by linearly combining base models from a simpler hypothesis class. In this work, we investigate the problem of adapting batch gradient boosting for minimizing convex loss functions to online setting where the loss at each iteration is i.i.d sampled from an unknown distribution. To generalize from batch to online, we ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Bernoulli
سال: 2021
ISSN: ['1573-9759', '1350-7265']
DOI: https://doi.org/10.3150/19-bej1187